Goto

Collaborating Authors

 main entity


Seemingly Plausible Distractors in Multi-Hop Reasoning: Are Large Language Models Attentive Readers?

Bhuiya, Neeladri, Schlegel, Viktor, Winkler, Stefan

arXiv.org Artificial Intelligence

State-of-the-art Large Language Models (LLMs) are accredited with an increasing number of different capabilities, ranging from reading comprehension, over advanced mathematical and reasoning skills to possessing scientific knowledge. In this paper we focus on their multi-hop reasoning capability: the ability to identify and integrate information from multiple textual sources. Given the concerns with the presence of simplifying cues in existing multi-hop reasoning benchmarks, which allow models to circumvent the reasoning requirement, we set out to investigate, whether LLMs are prone to exploiting such simplifying cues. We find evidence that they indeed circumvent the requirement to perform multi-hop reasoning, but they do so in more subtle ways than what was reported about their fine-tuned pre-trained language model (PLM) predecessors. Motivated by this finding, we propose a challenging multi-hop reasoning benchmark, by generating seemingly plausible multi-hop reasoning chains, which ultimately lead to incorrect answers. We evaluate multiple open and proprietary state-of-the-art LLMs, and find that their performance to perform multi-hop reasoning is affected, as indicated by up to 45% relative decrease in F1 score when presented with such seemingly plausible alternatives. We conduct a deeper analysis and find evidence that while LLMs tend to ignore misleading lexical cues, misleading reasoning paths indeed present a significant challenge.


Cisco MindMeld Chatbot Development Framework

#artificialintelligence

A word or phrase that provides information necessary to fulfill a particular intent. Each entity belongs to a category specified by the entity's associated type. For instance, a book_flight intent could have a location type for entities like'Miami' and'Chicago O'Hare', an airline type for entities like'Air India' and'Southwest', and a date type for entities like'July 4th' and'New Years Day'. An application-agnostic entity that is automatically detected by MindMeld. Examples include numbers, time expressions, email addresses, URLs and measured quantities like distance, volume, currency and temperature.


Author's Sentiment Prediction

Bastan, Mohaddeseh, Koupaee, Mahnaz, Son, Youngseo, Sicoli, Richard, Balasubramanian, Niranjan

arXiv.org Artificial Intelligence

We introduce PerSenT, a dataset of crowd-sourced annotations of the sentiment expressed by the authors towards the main entities in news articles. The dataset also includes paragraph-level sentiment annotations to provide more fine-grained supervision for the task. Our benchmarks of multiple strong baselines show that this is a difficult classification task. The results also suggest that simply fine-tuning document-level representations from BERT isn't adequate for this task. Making paragraph-level decisions and aggregating them over the entire document is also ineffective. We present empirical and qualitative analyses that illustrate the specific challenges posed by this dataset. We release this dataset with 5.3k documents and 38k paragraphs covering 3.2k unique entities as a challenge in entity sentiment analysis.


How AI will pass its driving test - Automotive World

#artificialintelligence

Artificial intelligence is learning to drive. Within the next 20 years it will take over from humans as the main entity behind the wheel. However, AI in the automotive industry is more than the technology merely passing its driving test. This piece explores not only the impact of AI on driving, but how it will fundamentally alter the way OEMs do business and the laws surrounding road use across the world. Manufacturers are sitting on a huge amount of information from cars already on the road, but are unable to harness it for sales purposes with insights buried amongst mountains of data. In thinking about AI in the automotive industry, there is one trend that can't be ignored: autonomous vehicles.


Law of Connectivity in Machine Learning

Dundas, Jitesh

arXiv.org Artificial Intelligence

We present in this paper our law that there is always a connection present between two entities, with a selfconnection being present at least in each node. An entity is an object, physical or imaginary, that is connected by a path (or connection) and which is important for achieving the desired result of the scenario. In machine learning, we state that for any scenario, a subject entity is always, directly or indirectly, connected and affected by single or multiple independent / dependent entities, and their impact on the subject entity is dependent on various factors falling into the categories such as the existenc